Fast implementations of nearest neighbor classifiers

نویسندگان

  • Patrick Grother
  • Gerald T. Candela
  • James L. Blue
چکیده

Standard implementations of non-parametric classifiers have large computational requirements. Parzen classifiers use the distances of an unknown vector to all N prototype samples, and consequently exhibit O(N) behavior in both memory and time. We describe four techniques for expediting the nearest neighbor methods. replacing the linear search with a new kd tree method, exhibiting approximately O(N) behavior; employing an L∞ instead of L2 distance metric; using variance ordered features; and rejecting prototypes by evaluating distances in low dimensionality subspaces. We demonstrate that varianceordered features yield significant efficiency gains over the same features linearly transformed to have uniform variance. We qive results for a large OCR problem, but note that the techniques expedite recognition for arbitrary applications. Three of four techniques preserve recognition accuracy.

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Simple and Fast Nearest Neighbor Search

We present a simple randomized data structure for twodimensional point sets that allows fast nearest neighbor queries in many cases. An implementation outperforms several previous implementations for commonly used benchmarks.

متن کامل

Fast Most Similar Neighbor (MSN) classifiers for Mixed Data

The k nearest neighbor (k-NN) classifier has been extensively used in Pattern Recognition because of its simplicity and its good performance. However, in large datasets applications, the exhaustive k-NN classifier becomes impractical. Therefore, many fast kNN classifiers have been developed; most of them rely on metric properties (usually the triangle inequality) to reduce the number of prototy...

متن کامل

Fusion of multiple approximate nearest neighbor classifiers for fast and efficient classification

The nearest neighbor classifier (NNC) is a popular non-parametric classifier. It is a simple classifier with no design phase and shows good performance. Important factors affecting the efficiency and performance of NNC are (i) memory required to store the training set, (ii) classification time required to search the nearest neighbor of a given test pattern, and (iii) due to the curse of dimensi...

متن کامل

Fuzzy nearest neighbor algorithms: Taxonomy, experimental analysis and prospects

In recent years, many nearest neighbor algorithms based on fuzzy sets theory have been developed. These methods form a field, known as fuzzy nearest neighbor classification, which is the source of many proposals for the enhancement of the k nearest neighbor classifier. Fuzzy sets theory and several extensions, including fuzzy rough sets, intuitionistic fuzzy sets, type-2 fuzzy sets and possibil...

متن کامل

Identification of selected monogeneans using image processing, artificial neural network and K-nearest neighbor

Abstract Over the last two decades, improvements in developing computational tools made significant contributions to the classification of biological specimens` images to their correspondence species. These days, identification of biological species is much easier for taxonomist and even non-taxonomists due to the development of automated computer techniques and systems.  In this study, we d...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:
  • Pattern Recognition

دوره 30  شماره 

صفحات  -

تاریخ انتشار 1997